156 research outputs found

    A Deep learning based food recognition system for lifelog images

    Get PDF
    In this paper, we propose a deep learning based system for food recognition from personal life archive im- ages. The system first identifies the eating moments based on multi-modal information, then tries to focus and enhance the food images available in these moments, and finally, exploits GoogleNet as the core of the learning process to recognise the food category of the images. Preliminary results, experimenting on the food recognition module of the proposed system, show that the proposed system achieves 95.97% classification accuracy on the food images taken from the personal life archive from several lifeloggers, which potentially can be extended and applied in broader scenarios and for different types of food categories

    Hybrid Transformer Network for Deepfake Detection

    Get PDF
    Deepfake media is becoming widespread nowadays because of the easily available tools and mobile apps which can generate realistic looking deepfake videos/images without requiring any technical knowledge. With further advances in this field of technology in the near future, the quantity and quality of deepfake media is also expected to flourish, while making deepfake media a likely new practical tool to spread mis/disinformation. Because of these concerns, the deepfake media detection tools are becoming a necessity. In this study, we propose a novel hybrid transformer network utilizing early feature fusion strategy for deepfake video detection. Our model employs two different CNN networks, i.e., (1) XceptionNet and (2) EfficientNet-B4 as feature extractors. We train both feature extractors along with the transformer in an end-to-end manner on FaceForensics++, DFDC benchmarks. Our model, while having relatively straightforward architecture, achieves comparable results to other more advanced state-of-the-art approaches when evaluated on FaceForensics++ and DFDC benchmarks. Besides this, we also propose novel face cut-out augmentations, as well as random cut-out augmentations. We show that the proposed augmentations improve the detection performance of our model and reduce overfitting. In addition to that, we show that our model is capable of learning from considerably small amount of data.publishedVersio

    Detecting Out-of-Context Image-Caption Pairs in News: A Counter-Intuitive Method

    Full text link
    The growth of misinformation and re-contextualized media in social media and news leads to an increasing need for fact-checking methods. Concurrently, the advancement in generative models makes cheapfakes and deepfakes both easier to make and harder to detect. In this paper, we present a novel approach using generative image models to our advantage for detecting Out-of-Context (OOC) use of images-caption pairs in news. We present two new datasets with a total of 68006800 images generated using two different generative models including (1) DALL-E 2, and (2) Stable-Diffusion. We are confident that the method proposed in this paper can further research on generative models in the field of cheapfake detection, and that the resulting datasets can be used to train and evaluate new models aimed at detecting cheapfakes. We run a preliminary qualitative and quantitative analysis to evaluate the performance of each image generation model for this task, and evaluate a handful of methods for computing image similarity.Comment: ACM International Conference on Content-Based Multimedia Indexing (CBMI '23

    A usage analytics model for analysing user behaviour in IBM academic cloud

    Get PDF
    Usage in the software domain refers to the knowledge about how end- users use the application and how the application responds to the user’s actions. Usage can be revealed by monitoring the user’s interaction with the application. However, in the cloud environment, it is non-trivial to understand the interactions of the users by using only the monitoring solutions. For example, user’s behaviour, user’s usage pattern, which features of a cloud application are critical for a user, to name a few, cannot be extracted using only the existing monitoring tools. Understanding these information require additional analysis, which can be done by using usage analytics. For this purpose, in this paper, we propose a novel process model design for incorporating Usage Analytics in a cloud environment. We evaluate this proposed process model in the context of academic applications and services in the cloud, with the focus on IBM Academic Cloud

    DCU at the NTCIR-13 Lifelog-2 Task

    Get PDF
    In this work, we outline the submissions of Dublin City University (DCU) team, the organisers, to the NTCIR-13 Lifelog-2 Task. We submitted runs to the Lifelog Semantics Access (LSAT) and the Lifelog Insight (LIT) sub-tasks

    An Overview of User-level Usage Monitoring in Cloud Environment

    Get PDF
    Cloud computing monitors applications, virtual and physical resources to ensure performance capacity, workload management, optimize future application updates and so on. Current state-of-the-art monitoring solutions in the cloud focus on monitoring in application/service level, virtual and physical (infrastructure) level. While some of the researchers have identified the importance of monitoring users, there is still need for developing solutions, implementation and evaluation in this domain. In this paper, we propose a novel approach to extract end-user usage of cloud services from their interactions with the interfaces provided to access the services called User-level Usage Monitoring. We provide the principles necessary for the usage data extraction process and analyse existing cloud monitoring techniques from the identified principles. Understanding end-user usage patterns and behaviour can help developers and architects to assess how applications work and which features of the application are critical for the users

    Replay detection and multi-stream synchronization in CS:GO game streams using content-based Image retrieval and Image signature matching

    Get PDF
    In GameStory: The 2019 Video Game Analytics Challenge, two main tasks are nominated to solve in the challenge, which are replay detection - multi-stream synchronization, and game story summarization. In this paper, we propose a data-driven based approach to solve the first task: replay detection - multi-stream synchronization. Our solution aims to determine the replays which lie between two logo-transitional endpoints and synchronize them with their sources by extracting frames from videos, then applying image processing and retrieval remedies. In detail, we use the Bag of Visual Words approach to detect the logo-transitional endpoints, which contains multiple replays in between, then employ an Image Signature Matching algorithm for multi-stream synchronization and replay boundaries refinement. The best configuration of our proposed solution manages to achieve the second-highest scores in all evaluation metrics of the challenge

    A RESEARCH ON MULTI-OBJECTIVE OPTIMIZATION OF THE GRINDING PROCESS USING SEGMENTED GRINDING WHEEL BY TAGUCHI-DEAR METHOD

    Get PDF
    In this study, the mutil-objective optimization was applied for the surface grinding process of SAE420 steel. The aluminum oxide grinding wheels that were grooved by 15 grooves, 18 grooves, and 20 grooves were used in the experimental process. The Taguchi method was applied to design the experimental matrix. Four input parameters that were chosen for each experiment were the number of grooves in cylinder surface of grinding wheel, workpiece velocity, feed rate, and cutting depth. Four output parameters that were measured for each experimental were the machining surface roughness, the system vibrations in the three directions (X, Y, Z). The DEAR technique was applied to determine the values of the input parameters to obtaine the minimum values of machining surface roughness and vibrations in three directions. By using this technique, the optimum values of grinding wheel groove number, workpiece velocity, feed-rate, cutting depth were 18 grooves, 15 m/min, 2 mm/stroke, and 0.005 mm, respectively. The verified experimental was performed by using the optimum values of input parameters. The validation results of surface roughness and vibrations in X, Y, Z directions were 0.826 (”m), 0.531 (”m), 0.549 (”m), and 0. 646 (”m), respectively. These results were great improved in comparing to the normal experimental results. Taguchi method and DEAR technique can be applied to improve the quality of grinding surface and reduce the vibrations of the technology system to restrain the increasing of the cutting forces in the grinding process. Finally, the research direction was also proposed in this stud

    Overview of ImageCLEF lifelog 2017: lifelog retrieval and summarization

    Get PDF
    Despite the increasing number of successful related work- shops and panels, lifelogging has rarely been the subject of a rigorous comparative benchmarking exercise. Following the success of the new lifelog evaluation task at NTCIR-12, the first ImageCLEF 2017 LifeLog task aims to bring the attention of lifelogging to a wide audience and to promote research into some of the key challenges of the coming years. The ImageCLEF 2017 LifeLog task aims to be a comparative evaluation framework for information access and retrieval systems operating over personal lifelog data. Two subtasks were available to participants; all tasks use a single mixed modality data source from three lifeloggers for a period of about one month each. The data contains a large collection of wearable camera images, an XML description of the semantic locations, as well as the physical activities of the lifeloggers. Additional visual concept information was also provided by exploiting the Caffe CNN-based visual concept detector. For the two sub-tasks, 51 topics were chosen based on the real interests of the lifeloggers. In this first year three groups participated in the task, submitting 19 runs across all subtasks, and all participants also provided working notes papers. In general, the groups performance is very good across the tasks, and there are interesting insights into these very relevant challenges

    Organizer team at ImageCLEFlifelog 2017: baseline approaches for lifelog retrieval and summarization

    Get PDF
    This paper describes the participation of Organizer Team in the ImageCLEFlifelog 2017 Retrieval and Summarization subtasks. In this paper, we propose some baseline approaches, using only the provided information, which require different involvement levels from the users. With these baselines we target at providing references for other approaches that aim to solve the problems of lifelog retrieval and summarization
    • 

    corecore